8 research outputs found

    Symbolic Music Representations for Classification Tasks: A Systematic Evaluation

    Full text link
    Music Information Retrieval (MIR) has seen a recent surge in deep learning-based approaches, which often involve encoding symbolic music (i.e., music represented in terms of discrete note events) in an image-like or language like fashion. However, symbolic music is neither an image nor a sentence, and research in the symbolic domain lacks a comprehensive overview of the different available representations. In this paper, we investigate matrix (piano roll), sequence, and graph representations and their corresponding neural architectures, in combination with symbolic scores and performances on three piece-level classification tasks. We also introduce a novel graph representation for symbolic performances and explore the capability of graph representations in global classification tasks. Our systematic evaluation shows advantages and limitations of each input representation. Our results suggest that the graph representation, as the newest and least explored among the three approaches, exhibits promising performance, while being more light-weight in training

    8+8=4: Formalizing Time Units to Handle Symbolic Music Durations

    Full text link
    This paper focuses on the nominal durations of musical events (notes and rests) in a symbolic musical score, and on how to conveniently handle these in computer applications. We propose the usage of a temporal unit that is directly related to the graphical symbols in musical scores and pair this with a set of operations that cover typical computations in music applications. We formalize this time unit and the more commonly used approach in a single mathematical framework, as semirings, algebraic structures that enable an abstract description of algorithms/processing pipelines. We then discuss some practical use cases and highlight when our system can improve such pipelines by making them more efficient in terms of data type used and the number of computations.Comment: In Proceedings of the International Symposium on Computer Music Multidisciplinary Research (CMMR 2023), Tokyo, Japa

    The ACCompanion: Combining Reactivity, Robustness, and Musical Expressivity in an Automatic Piano Accompanist

    Full text link
    This paper introduces the ACCompanion, an expressive accompaniment system. Similarly to a musician who accompanies a soloist playing a given musical piece, our system can produce a human-like rendition of the accompaniment part that follows the soloist's choices in terms of tempo, dynamics, and articulation. The ACCompanion works in the symbolic domain, i.e., it needs a musical instrument capable of producing and playing MIDI data, with explicitly encoded onset, offset, and pitch for each played note. We describe the components that go into such a system, from real-time score following and prediction to expressive performance generation and online adaptation to the expressive choices of the human player. Based on our experience with repeated live demonstrations in front of various audiences, we offer an analysis of the challenges of combining these components into a system that is highly reactive and precise, while still a reliable musical partner, robust to possible performance errors and responsive to expressive variations.Comment: In Proceedings of the 32nd International Joint Conference on Artificial Intelligence (IJCAI-23), Macao, China. The differences/extensions with the previous version include a technical appendix, added missing links, and minor text updates. 10 pages, 4 figure

    Automatic Note-Level Score-to-Performance Alignments in the ASAP Dataset

    Get PDF
    Several MIR applications require fine-grained note alignments between MIDI performances and their musical scores for training and evaluation. However, large and high-quality datasets with this kind of data are not available, and their manual creation is a very time-consuming task that can only be performed by field experts. In this paper, we evaluate state-of-the-art automatic note alignment models applied to dataset generation. We increase the accuracy and reliability of the produced alignments with models that flexibly leverage existing annotations such as beat or measure alignments. We thoroughly evaluate these segment-constrained models and use the best to create note alignments for the ASAP dataset, a large dataset of solo piano MIDI performances beat-aligned to MusicXML scores. The resulting note alignments are manually checked and publicly available at: https://github.com/CPJKU/asap-dataset. The contributions of this paper are four-fold: (1) we extend the ASAP dataset with reliable note alignments, thus creating (n)ASAP, the largest available fully note-aligned dataset, comprising more than 7 M annotated notes and close to 100 hours of music; (2) we design, evaluate, and publish segment-constrained models for note alignments that flexibly leverage existing annotations and significantly outperform automatic models; (3) we design, evaluate, and publish unconstrained automatic models for note alignment that produce results on par with the state of the art; (4) we introduce Parangonada, a web-interface for visualizing and correcting alignment annotations

    Partitura: A Python Package for Symbolic Music Processing

    Get PDF
    Partitura is a lightweight Python package for handling symbolic musical information. It provides easy access to features commonly used in music information retrieval tasks, like note arrays (lists of timed pitched events) and 2D piano roll matrices, as well as other score elements such as time and key signatures, performance directives, and repeat structures. Partitura can load musical scores (in MEI, MusicXML, Humdrum **kern, and MIDI formats), MIDI performances, and score- to-performance alignments. The package includes some tools for music analysis, such as automatic pitch spelling, key signature identification, and voice separation. Partitura is an open-source project and is available at https://github.com/CPJKU/partitura/

    The Match File Format: Encoding Alignments Between Scores and Performances

    Get PDF
    This paper presents the specifications of match: a file format that extends a MIDI human performance with note-, beat-, and downbeat-level alignments to a corresponding musical score. This enables advanced analyses of the performance that are relevant for various tasks, such as expressive performance modelling, score following, music transcription, and performer classification. The match file includes a set of score-related descriptors that makes it usable also as a bare-bones score representation. For applications that require the use of structural score elements (e.g., voices, parts, beams, slurs), the match file can be easily combined with the symbolic score. To support the practical application of our work, we release a corrected and upgraded version of the Vienna 4x22 dataset of scores and performances aligned with match files

    MUSIC GENRE DESCRIPTOR FOR CLASSIFICATION BASED ON TONNETZ TRAJECTORIES

    No full text
    International audienceDans cet article, nous présentons un nouveau descripteur pour la classification automatique du style musical. Notre méthode consiste à définir une trajectoire harmonique dans un espace géométrique, le Tonnetz, puis à la résumer à ses valeurs de centralité, qui constituent les descripteurs. Ceux-ci, associés à des descripteurs classiques, sont utilisés comme caractéristiques pour la classification. Les résultats montrent des scores F 1 supérieurs à 0,8 avec une méthode classique de forêts aléatoires pour 8 classes (une par compositeur), et supérieurs à 0,9 pour une classification en 4 classes de style ou période de composition

    Musical genre descriptor for classification based on Tonnetz trajectories

    No full text
    National audienceDans cet article, nous présentons un nouveau descripteur pour la classification automatique du style musical. Notre méthode consiste à définir une trajectoire harmonique dans un espace géométrique, le Tonnetz, puis à la résumer à ses valeurs de centralité, qui constituent les descripteurs. Ceux-ci, associés à des descripteurs classiques, sont utilisés comme caractéristiques pour la classification. Les résultats montrent des scores F 1 supérieurs à 0,8 avec une méthode classique de forêts aléatoires pour 8 classes (une par compositeur), et supérieurs à 0,9 pour une classification en 4 classes de style ou période de composition
    corecore